SEQ2SEQ++: A Multitasking-Based Seq2seq Model to Generate Meaningful and Relevant Answers
نویسندگان
چکیده
Question-answering chatbots have tremendous potential to complement humans in various fields. They are implemented using either rule-based or machine learning-based systems. Unlike the former, more scalable. Sequence-to-sequence (Seq2Seq) learning is one of most popular approaches and has shown remarkable progress since its introduction 2014. However, based on Seq2Seq show a weakness that it tends generate answers can be generic inconsistent with questions, thereby becoming meaningless and, therefore, may lower chatbot adoption rate. This attributed three issues: question encoder overfit, answer generation language model influence. Several recent methods utilize multitask (MTL) address this weakness. existing MTL models very little improvement over single-task learning, wherein they still answers. paper presents novel approach for called SEQ2SEQ++, which comprises multifunctional encoder, an decoder, ternary classifier. Additionally, SEQ2SEQ++ utilizes dynamic tasks loss weight mechanism calculation attention comprehensive mechanism. Experiments NarrativeQA SQuAD datasets were conducted gauge performance proposed comparison two recently models. The experimental results yields noteworthy improvements bilingual evaluation understudy, word error rate, Distinct-2 metrics.
منابع مشابه
Seq2seq-Attention Question Answering Model
A sequence-to-sequence attention reading comprehension model was implemented to fulfill Question Answering task defined in Stanford Question Answering Dataset (SQuAD). The basic structure was bidirectional LSTM (BiLSTM) encodings with attention mechanism as well as BiLSTM decoding. Several adjustments such as dropout, learning rate decay, and gradients clipping were used. Finally, the model ach...
متن کاملStackSeq2Seq: Dual Encoder Seq2Seq Recurrent Networks
A widely studied non-deterministic polynomial time (NP) hard problem lies in nding a route between the two nodes of a graph. Oen meta-heuristics algorithms such asA∗ are employed on graphs with a large number of nodes. Here, we propose a deep recurrent neural network architecture based on the Sequence-2-Sequence (Seq2Seq) model, widely used, for instance in text translation. Particularly, we ...
متن کاملTable-to-text Generation by Structure-aware Seq2seq Learning
Table-to-text generation aims to generate a description for a factual table which can be viewed as a set of field-value records. To encode both the content and the structure of a table, we propose a novel structure-aware seq2seq architecture which consists of field-gating encoder and description generator with dual attention. In the encoding phase, we update the cell memory of the LSTM unit by ...
متن کاملCold Fusion: Training Seq2Seq Models Together with Language Models
Sequence-to-sequence (Seq2Seq) models with attention have excelled at tasks which involve generating natural language sentences such as machine translation, image captioning and speech recognition. Performance has further been improved by leveraging unlabeled data, often in the form of a language model. In this work, we present the Cold Fusion method, which leverages a pre-trained language mode...
متن کاملSeq2seq for Morphological Reinflection: When Deep Learning Fails
Recent studies showed that the sequenceto-sequence (seq2seq) model is a promising approach for morphological reinflection. At the CoNLL-SIGMORPHON 2017 Shared Task for universal morphological reinflection, we basically followed the approach with some minor variations. The results were remarkable in a certain sense. In high-resource scenarios our system achieved 91.46% accuracy (only modestly be...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2021
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2021.3133495